The Equivalence between Orthogonal Iterations and Alternating Least Squares

نویسندگان

چکیده

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Displacement Preconditioner for Toeplitz Least Squares Iterations

We consider the solution of least squares problems min ||b − Ax||2 by the preconditioned conjugate gradient (PCG) method, for m × n complex Toeplitz matrices A of rank n. A circulant preconditioner C is derived using the T. Chan optimal preconditioner for n × n matrices using the displacement representation of AA. This allows the fast Fourier transform (FFT) to be used throughout the computatio...

متن کامل

Accurate principal component analysis via a few iterations of alternating least squares

A few iterations of alternating least squares with a random starting point provably suffice to produce nearly optimal spectraland Frobenius-norm accuracies of low-rank approximations to a matrix; iterating to convergence is unnecessary. Thus, software implementing alternating least squares can be retrofitted via appropriate setting of parameters to calculate nearly optimally accurate low-rank a...

متن کامل

On the Difference Between Orthogonal Matching Pursuit and Orthogonal Least Squares

Greedy algorithms are often used to solve underdetermined inverse problems when the solution is constrained to be sparse, i.e. the solution is only expected to have a relatively small number of non-zero elements. Two different algorithms have been suggested to solve such problems in the signal processing and control community, orthogonal Matching Pursuit and orthogonal Least Squares respectivel...

متن کامل

Vector Orthogonal Polynomials and Least Squares Approximation

We describe an algorithm for complex discrete least squares approximation, which turns out to be very efficient when function values are prescribed in points on the real axis or on the unit circle. In the case of polynomial approximation, this reduces to algorithms proposed by Rutishauser, Gragg, Harrod, Reichel, Ammar and others. The underlying reason for efficiency is the existence of a recur...

متن کامل

Orthogonal Low Rank Tensor Approximation: Alternating Least Squares Method and Its Global Convergence

With the notable exceptions of two cases — that tensors of order 2, namely, matrices, always have best approximations of arbitrary low ranks and that tensors of any order always have the best rank-one approximation, it is known that high-order tensors may fail to have best low rank approximations. When the condition of orthogonality is imposed, even under the modest assumption that only one set...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Advances in Linear Algebra & Matrix Theory

سال: 2020

ISSN: 2165-333X,2165-3348

DOI: 10.4236/alamt.2020.102002